About
In advancing Visual Question Answering (VQA) models, I utilized innovative methodologies. This involved generating a new dataset and designing interpretable VQA models specifically for the Semi-CLEVR dataset. Leveraging large language models (LLMs), causal explanations were developed to enhance model interpretability. Additionally, diffusion models were applied to the novel CLEVR dataset to accurately decompose occluded shapes, capitalizing on their capacity to model complex data distributions and uncover hidden details. This research aimed to enhance generalization in VQA through the integration of causal reasoning and advanced modeling techniques.
- Programming Languages: Python, C++
- Databases: MySQL, MongoDB, PostgreSQL
- Libraries & Frameworks: PyTorch, TensorFlow, Hugging Face Transformers, Keras, NumPy, Pandas, OpenCV
- Model Optimization & Deployment: ONNX, TensorRT, TorchServe, FastAPI
- Tools & Platforms: Git, Docker, Kubernetes, AWS, GCP, Azure, JIRA, Weights & Biases (wandb)
Seeking a challenging position that leverages my expertise in Machine Learning and Software Engineering, offering opportunities for professional development, innovative experiences, and personal growth.
Experience
- Fine-tuned LLM models, including Mistral, LLaMA, and GPT, leveraging techniques such as activation sparsity and attention sparsity to optimize performance.
- Applied techniques such as KWTA, dynamic context pruning, and KV caching to enhance model efficiency.
- Tools: Python, PyTorch, Accelerate, GPT, llama
- During my internship at Petrolern as a Digital Innovation Intern, I gained experience in both machine learning and data compression techniques
- I developed a semantic compression technique using a deep autoencoder to effectively map data tuples into a lower-dimensional representation
- As a machine learning engineer, I built models for analyzing geothermal data and improved their performance through algorithmic optimization
- Fine-tuned models like BART for summarization on Persian text data.
- Implemented Matrix Factorization for topic modeling.
- utilized BiLSTM-CRF Models for sequential tagging.
- Tools: Python, Scikit-learn, NLTK
Projects
SHAP vs Lime Vs ELI5
- Tools: Python, PyTorch
- To explain the model's predictions, the project uses model interpretability tools such as SHAP (SHapley Additive exPlanations), Lime (Local Interpretable Model-agnostic Explanations), and Eli5 (Explain Like I'm 5). These tools provide insights into how the model makes decisions and highlight the importance of different features in predicting strokes..
Skills
Languages and Databases
Libraries
Frameworks
Other
Education
Laramie, WY
Degree: PhD in Computer Science
Area of Study: Causal Reasoning for Improving Generalization in Visual Question Answering
- Intro to Artificial Intelligen
- Machine Learning
- High Perform Comput & Paradigm
- Advanced Image Processing
- Neural and Fuzzy Systems
Relevant Courseworks:
Tehran, Iran
Degree: Masters of Information Technology
CGPA: 3.68/4.0
- Artificial Neural Networks
- Neural and Fuzzy Systems
Relevant Courseworks: